DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5...9
Hits 1 – 20 of 167

1
Europarl Direct Translationese Dataset ...
BASE
Show details
2
Europarl Direct Translationese Dataset ...
BASE
Show details
3
Europarl Direct Translationese Dataset ...
BASE
Show details
4
Integrating Unsupervised Data Generation into Self-Supervised Neural Machine Translation for Low-Resource Languages ...
BASE
Show details
5
Comparing Feature-Engineering and Feature-Learning Approaches for Multilingual Translationese Classification ...
BASE
Show details
6
Investigating the Helpfulness of Word-Level Quality Estimation for Post-Editing Machine Translation Output ...
BASE
Show details
7
Multi-Head Highly Parallelized LSTM Decoder for Neural Machine Translation ...
BASE
Show details
8
Comparing Feature-Engineering and Feature-Learning Approaches for Multilingual Translationese Classification ...
BASE
Show details
9
Modeling Task-Aware MIMO Cardinality for Efficient Multilingual Neural Machine Translation ...
BASE
Show details
10
A Bidirectional Transformer Based Alignment Model for Unsupervised Word Alignment ...
BASE
Show details
11
Automatic classification of human translation and machine translation : a study from the perspective of lexical diversity
Fu, Yingxue; Nederhof, Mark Jan. - : Linkoping University Electronic Press, 2021
BASE
Show details
12
Transformer-based NMT : modeling, training and implementation
Xu, Hongfei. - : Saarländische Universitäts- und Landesbibliothek, 2021
Abstract: International trade and industrial collaborations enable countries and regions to concentrate their developments on specific industries while making the most of other countries' specializations, which significantly accelerates global development. However, globalization also increases the demand for cross-region communication. Language barriers between many languages worldwide create a challenge for achieving deep collaboration between groups speaking different languages, increasing the need for translation. Language technology, specifically, Machine Translation (MT) holds the promise to enable communication between languages efficiently in real-time with minimal costs. Even though nowadays computers can perform computation in parallel very fast, which provides machine translation users with translations with very low latency, and although the evolution from Statistical Machine Translation (SMT) to Neural Machine Translation (NMT) with the utilization of advanced deep learning algorithms has significantly boosted translation quality, current machine translation algorithms are still far from accurately translating all input. Thus, how to further improve the performance of state-of-the-art NMT algorithm remains a valuable open research question which has received a wide range of attention. In the research presented in this thesis, we first investigate the long-distance relation modeling ability of the state-of-the-art NMT model, the Transformer. We propose to learn source phrase representations and incorporate them into the Transformer translation model, aiming to enhance its ability to capture long-distance dependencies well. Second, though previous work (Bapna et al., 2018) suggests that deep Transformers have difficulty in converging, we empirically find that the convergence of deep Transformers depends on the interaction between the layer normalization and residual connections employed to stabilize its training. We conduct a theoretical study about how to ensure the convergence of Transformers, especially for deep Transformers, and propose to ensure the convergence of deep Transformers by putting the Lipschitz constraint on its parameter initialization. Finally, we investigate how to dynamically determine proper and efficient batch sizes during the training of the Transformer model. We find that the gradient direction gets stabilized with increasing batch size during gradient accumulation. Thus we propose to dynamically adjust batch sizes during training by monitoring the gradient direction change within gradient accumulation, and to achieve a proper and efficient batch size by stopping the gradient accumulation when the gradient direction starts to fluctuate. For our research in this thesis, we also implement our own NMT toolkit, the Neutron implementation of the Transformer and its variants. In addition to providing fundamental features as the basis of our implementations for the approaches presented in this thesis, we support many advanced features from recent cutting-edge research work. Implementations of all our approaches in this thesis are also included and open-sourced in the toolkit. To compare with previous approaches, we mainly conducted our experiments on the data from the WMT 14 English to German (En-De) and English to French (En-Fr) news translation tasks, except when studying the convergence of deep Transformers, where we alternated the WMT 14 En-Fr task with the WMT 15 Czech to English (Cs-En) news translation task to compare with Bapna et al. (2018). The sizes of these datasets vary from medium (the WMT 14 En-De, ~ 4.5M sentence pairs) to very large (the WMT 14 En-Fr, ~ 36M sentence pairs), thus we suggest our approaches help improve the translation quality between popular language pairs which are widely used and have sufficient data. ; China Scholarship Council
Keyword: ddc:430; ddc:440; ddc:490; ddc:491.8; ddc:600; dynamic batch size; neural machine translation; optimization; parameter initialization; phrase representation; transformer translation model
URL: https://doi.org/10.22028/D291-34998
http://nbn-resolving.org/urn:nbn:de:bsz:291--ds-349988
BASE
Hide details
13
The European Language Technology Landscape in 2020: Language-Centric and Human-Centric AI for Cross-Cultural Communication in Multilingual Europe
In: Language Resources and Evaluation Conference ; https://hal.archives-ouvertes.fr/hal-02892154 ; Language Resources and Evaluation Conference, ELDA/ELRA, May 2020, Marseille, France ; https://lrec2020.lrec-conf.org/en/ (2020)
BASE
Show details
14
The European Language Technology Landscape in 2020: Language-Centric and Human-Centric AI for Cross-Cultural Communication in Multilingual Europe ...
BASE
Show details
15
The European Language Technology Landscape in 2020: Language-Centric and Human-Centric AI for Cross-Cultural Communication in Multilingual Europe ...
BASE
Show details
16
The European Language Technology Landscape in 2020: Language-Centric and Human-Centric AI for Cross-Cultural Communication in Multilingual Europe ...
BASE
Show details
17
Linguistically inspired morphological inflection with a sequence to sequence model ...
BASE
Show details
18
Probing Word Translations in the Transformer and Trading Decoder for Encoder Layers ...
BASE
Show details
19
Language service provision in the 21st century: challenges, opportunities and educational perspectives for translation studies
In: ISBN: 9788869234934 ; Bologna Process beyond 2020: Fundamental values of the EHEA pp. 297-303 (2020)
BASE
Show details
20
Deep interactive text prediction and quality estimation in translation interfaces
Hokamp, Christopher M.. - : Dublin City University. School of Computing, 2018
In: Hokamp, Christopher M. (2018) Deep interactive text prediction and quality estimation in translation interfaces. PhD thesis, Dublin City University. (2018)
BASE
Show details

Page: 1 2 3 4 5...9

Catalogues
0
0
1
0
5
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
161
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern